Dynamic Deep Neural Networks: Optimizing Accuracy-Efficiency Trade-offs by Selective Execution

نویسندگان

  • Lanlan Liu
  • Jia Deng
چکیده

We introduce Dynamic Deep Neural Networks (DNN), a new type of feed-forward deep neural network that allows selective execution. Given an input, only a subset of DNN neurons are executed, and the particular subset is determined by the DNN itself. By pruning unnecessary computation depending on input, DNNs provide a way to improve computational efficiency. To achieve dynamic selective execution, a DNN augments a feed-forward deep neural network (directed acyclic graph of differentiable modules) with controller modules. Each controller module is a sub-network whose output is a decision that controls whether other modules can execute. A DNN is trained end to end. Both regular and controller modules in a DNN are learnable and are jointly trained to optimize both accuracy and efficiency. Such training is achieved by integrating backpropagation with reinforcement learning. With extensive experiments of various DNN architectures on image classification tasks, we demonstrate that DNNs are general and flexible, and can effectively optimize accuracyefficiency trade-offs.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

GradNets: Dynamic Interpolation Between Neural Architectures

In machine learning, there is a fundamental trade-off between ease of optimization and expressive power. Neural Networks, in particular, have enormous expressive power and yet are notoriously challenging to train. The nature of that optimization challenge changes over the course of learning. Traditionally in deep learning, one makes a static trade-off between the needs of early and late optimiz...

متن کامل

Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability

Tartan (TRT), a hardware accelerator for inference with Deep Neural Networks (DNNs), is presented and evaluated on Convolutional Neural Networks. TRT exploits the variable per layer precision requirements of DNNs to deliver execution time that is proportional to the precision p in bits used per layer for convolutional and fully-connected layers. Prior art has demonstrated an accelerator with th...

متن کامل

A Dversarial R Obustness of P Runed N Eural N Etworks

Deep neural network pruning forms a compressed network by discarding “unimportant” weights or filters. Standard evaluation metrics have shown their remarkable speedup and prediction accuracy in test time, but their adversarial robustness remains unexplored even though it is an important security feature in deployment. We study the robustness of pruned neural networks under adversarial attacks. ...

متن کامل

Compressive ELM: Improved Models through Exploiting Time-Accuracy Trade-Offs

In the training of neural networks, there often exists a tradeoff between the time spent optimizing the model under investigation, and its final performance. Ideally, an optimization algorithm finds the model that has best test accuracy from the hypothesis space as fast as possible, and this model is efficient to evaluate at test time as well. However, in practice, there exists a trade-off betw...

متن کامل

Optimizing the Event-based Method of Localization in Wireless Sensor Networks

A Wireless Sensor Network (WSN) is a wireless decentralized structure network consists of many nodes. Nodes can be fixed or mobile. WSN applications typically observe some physical phenomenon through sampling of the environment so determine the location of events is an important issue in WSN. Wireless Localization used to determine the position of nodes. The precise localization in WSNs is a co...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1701.00299  شماره 

صفحات  -

تاریخ انتشار 2017